Goto

Collaborating Authors

 weak learner





Language models are weak learners

Neural Information Processing Systems

A central notion in practical and theoretical machine learning is that of a weak learner, classifiers that achieve better-than-random performance (on any given distribution over data), even by a small margin. Such weak learners form the practical basis for canonical machine learning methods such as boosting.






daf8364f0715a41a469c677c0adc4754-Supplemental-Conference.pdf

Neural Information Processing Systems

Since weak learners perform only marginallybetter than random guesses, such subroutines constitute aweakerassumption than the availability of an accurate supervised learning oracle. Weprovethat the sample complexity and running time bounds of the proposed method do not explicitly dependonthenumberofstates. While existing results on boosting operate on convex losses, the value function over policies is non-convex.